276 research outputs found

    A deep learning model for the analysis of medical reports in ICD-10 clinical coding task

    Get PDF
    The practice of assigning a uniquely identifiable and easily traceable code to pathology from medical diagnoses is an added value to the current modality of archiving health data collected to build the clinical history of each of us. Unfortunately, the enormous amount of possible pathologies and medical conditions has led to the realization of extremely wide international codifications that are difficult to consult even for a human being. This difficulty makes the practice of annotation of diagnoses with ICD-10 codes very cumbersome and rarely performed. In order to support this operation, a classification model was proposed, able to analyze medical diagnoses written in natural language and automatically assign one or more international reference codes. The model has been evaluated on a dataset released in the Spanish language for the eHealth challenge (CodiEsp) of the international conference CLEF 2020, but it could be extended to any language with latin characters. We proposed a model based on a two-step classification process based on BERT and BiLSTM. Although still far from an accuracy sufficient to do without a licensed physician opinion, the results obtained show the feasibility of the task and are a starting point for future studies in this direction

    SciRecSys: A Recommendation System for Scientific Publication by Discovering Keyword Relationships

    Full text link
    In this work, we propose a new approach for discovering various relationships among keywords over the scientific publications based on a Markov Chain model. It is an important problem since keywords are the basic elements for representing abstract objects such as documents, user profiles, topics and many things else. Our model is very effective since it combines four important factors in scientific publications: content, publicity, impact and randomness. Particularly, a recommendation system (called SciRecSys) has been presented to support users to efficiently find out relevant articles

    Towards a social robot as interface for tourism recommendations

    Get PDF
    The popularity of social robots is steadily increasing, mainly due to the interesting impact they have in several application domains. In this paper, we propose the use of Pepper Robot as an interface of a recommender system for tourism. In particular, we used the robot to interact with the users and to provide them with personalized recommendations about hotels, restaurants, and points of interest in the area. The personalization mechanism encoded in the social robot relies on soft biometrics traits automatically recognized by the robot, as age and gender, user interests and personal facets. All these data are used to feed a neural network that returns as output the most suitable recommendations for the target user. To evaluate the effectiveness of the interaction driven by a social robot, we carried out a user study whose goal was to evaluate: (1) how the robot affects the perceived accuracy of the recommendations; (2) how the user experience and the engagement vary by interacting with a social robot instead of a classic web application. Even if there is a large room for improvement, mainly due to the poor speech recognizer integrated in the Pepper, the results showed that the robot can strongly attract people, thanks to its presence and interaction capabilities. These findings encouraged us in performing a larger field study to test the approach in the wild and to understand whether it can increase the acceptance of recommendations in real environments

    Extracting relations from Italian wikipedia using self-training

    Get PDF
    In this paper, we describe a supervised approach for extracting relations from Wikipedia. In particular, we exploit a self-training strategy for enriching a small number of manually labeled triples with new self-labeled examples. We integrate the supervised stage in WikiOIE, an existing framework for unsupervised extraction of relations from Wikipedia. We rely on WikiOIE and its unsupervised pipeline for extracting the initial set of unlabelled triples. An evaluation involving different algorithms and parameters proves that self-training helps to improve performance. Finally, we provide a dataset of about three million triples extracted from the Italian version of Wikipedia and perform a preliminary evaluation conducted on a sample dataset, obtaining promising results

    A domain-independent framework for building conversational recommender systems

    Get PDF
    Conversational Recommender Systems (CoRSs) implement a paradigm where users can interact with the system for defining their preferences and discovering items that best fit their needs. A CoRS can be straightforwardly implemented as a chatbot. Chatbots are becoming more and more popular for several applications like customer care, health care, medical diagnoses. In the most complex form, the implementation of a chatbot is a challenging task since it requires knowledge about natural language processing, human-computer interaction, and so on. In this paper, we propose a general framework for making easy the generation of conversational recommender systems. The framework, based on a content-based recommendation algorithm, is independent from the domain. Indeed, it allows to build a conversational recommender system with different interaction modes (natural language, buttons, hybrid) for any domain. The framework has been evaluated on two state-of-the-art datasets with the aim of identifying the components that mainly influence the final recommendation accuracy

    A study of Machine Learning models for Clinical Coding of Medical Reports at CodiEsp 2020

    Get PDF
    The task of identifying one or more diseases associated with a patient’s clinical condition is often very complex, even for doctors and specialists. This process is usually time-consuming and has to take into account different aspects of what has occurred, including symptoms elicited and previous healthcare situations. The medical diagnosis is often provided to patients in the form of written paper without any correlation with a national or international standard. Even if the WHO (World Health Organization) released the ICD10 international glossary of diseases, almost no doctor has enough time to manually associate the patient’s clinical history with international codes. The CodiEsp task at CLEF 2020 addressed this issue by proposing the development of an automatic system to deal with this task. Our solution investigated different machine learning strategies in order to identify an approach to face that challenge. The main outcomes of the experiments showed that a strategy based on BERT for pre-filtering and one based on BiLSTMCNN-SelfAttention for classification provide valuable results. We carried out several experiments on a subset of the training set for tuning the final model submitted to the challenge. In particular, we analyzed the impact of the algorithm, the input encoding strategy, and the thresholds for multi-label classification. A set of experiments has been carried out also during a post hoc analysis. The experiments confirmed that the strategy submitted to the CodiEsp task is the best performing one among those evaluated, and it allowed us to obtain a final mean average error value on the test set equal to 0.202. To support future developments of the proposed approach and the replicability of the experiments we decided to make the source code publicly accessible

    Research Paper Recommender System with Serendipity Using Tweets vs. Diversification

    Get PDF
    21st International Conference on Asia-Pacific Digital Libraries, ICADL 2019, Kuala Lumpur, Malaysia, November 4–7, 2019. Part of the Lecture Notes in Computer Science book series (LNCS, volume 11853), also part of the Information Systems and Applications, incl. Internet/Web, and HCI book sub series (LNISA, volume 11853).So far, a lot of works have studied research paper recommender systems. However, most of them have focused only on the accuracy and ignored the serendipity, which is an important aspect for user satisfaction. The serendipity is concerned with the novelty of recommendations and to which extent recommendations positively surprise users. In this paper, we investigate a research paper recommender system focusing on serendipity. In particular, we examine (1) whether a user’s tweets lead to a generation of serendipitous recommendations and (2) whether the use of diversification on a recommendation list improves serendipity. We have conducted an online experiment with 22 subjects in the domain of computer science. The result of our experiment shows that tweets do not improve the serendipity, despite their heterogeneous nature. However, diversification delivers serendipitous research papers that cannot be generated by a traditional strategy

    A personalized and context-aware news offer for mobile devices

    Get PDF
    For classical domains, such as movies, recommender systems have proven their usefulness. But recommending news is more challenging due to the short life span of news content and the demand for up-to-date recommendations. This paper presents a news recommendation service with a content-based algorithm that uses features of a search engine for content processing and indexing, and a collaborative filtering algorithm for serendipity. The extension towards a context-aware algorithm is made to assess the information value of context in a mobile environment through a user study. Analyzing interaction behavior and feedback of users on three recommendation approaches shows that interaction with the content is crucial input for user modeling. Context-aware recommendations using time and device type as context data outperform traditional recommendations with an accuracy gain dependent on the contextual situation. These findings demonstrate that the user experience of news services can be improved by a personalized context-aware news offer

    HPLC-HRMS Global Metabolomics Approach for the Diagnosis of “Olive Quick Decline Syndrome” Markers in Olive Trees Leaves

    Get PDF
    Olive quick decline syndrome (OQDS) is a multifactorial disease affecting olive plants. The onset of this economically devastating disease has been associated with a Gram-negative plant pathogen called Xylella fastidiosa (Xf). Liquid chromatography separation coupled to high-resolution mass spectrometry detection is one the most widely applied technologies in metabolomics, as it provides a blend of rapid, sensitive, and selective qualitative and quantitative analyses with the ability to identify metabolites. The purpose of this work is the development of a global metabolomics mass spectrometry assay able to identify OQDS molecular markers that could discriminate between healthy (HP) and infected (OP) olive tree leaves. Results obtained via multivariate analysis through an HPLC-ESI HRMS platform (LTQ-Orbitrap from Thermo Scientific) show a clear separation between HP and OP samples. Among the differentially expressed metabolites, 18 different organic compounds highly expressed in the OP group were annotated; results obtained by this metabolomic approach could be used as a fast and reliable method for the biochemical characterization of OQDS and to develop targeted MS approaches for OQDS detection by foliage analysis

    Automatic Discovery of Complementary Learning Resources

    Get PDF
    Proceedings of: 6th European Conference of Technology Enhanced Learning, EC-TEL 2011, Palermo, Italy, September 20-23, 2011.Students in a learning experience can be seen as a community working simultaneously (and in some cases collaboratively) in a set of activities. During these working sessions, students carry out numerous actions that affect their learning. But those actions happening outside a class or the Learning Management System cannot be easily observed. This paper presents a technique to widen the observability of these actions. The set of documents browsed by the students in a course was recorded during a period of eight weeks. These documents are then processed and the set with highest similarity with the course notes are selected and recommended back to all the students. The main problem is that this user community visits thousands of documents and only a small percent of them are suitable for recommendation. Using a combination of lexican analysis and information retrieval techniques, a fully automatic procedure to analyze these documents, classify them and select the most relevant ones is presented. The approach has been validated with an empirical study in an undergraduate engineering course with more than one hundred students. The recommended resources were rated as "relevant to the course" by the seven instructors with teaching duties in the course.Work partially funded by the Learn3 project, “Plan Nacional de I+D+I TIN2008-05163/TSI”, the Acción Integrada Ref. DE2009-0051, the “Emadrid: Investigación y desarrollo de tecnologías para el e-learning en la Comunidad de Madrid” project (S2009/TIC-1650) and TELMA Project (Plan Avanza TSI-020110-2009-85)
    corecore